Scalable Metric Learning via Weighted Approximate Rank Component Analysis
نویسندگان
چکیده
Our goal is to learn a Mahalanobis distance by minimizing a loss defined on the weighted sum of the precision at different ranks. Our core motivation is that minimizing a weighted rank loss is a natural criterion for many problems in computer vision such as person re-identification. We propose a novel metric learning formulation called Weighted Approximate Rank Component Analysis (WARCA). We then derive a scalable stochastic gradient descent algorithm for the resulting learning problem. We also derive an efficient non-linear extension of WARCA by using the kernel trick. Kernel space embedding decouples the training and prediction costs from the data dimension and enables us to plug inarbitrary distance measures which are more natural for the features. We also address a more general problem of matrix rank degeneration & non-isolated minima in the low-rank matrix optimization by using new type of regularizer which approximately enforces the orthonormality of the learned matrix very efficiently. We validate this new method on nine standard person re-identification datasets including two large scale Market1501 and CUHK03 datasets and show that we improve upon the current state-of-the-art methods on all of them.
منابع مشابه
WMRB: Learning to Rank in a Scalable Batch Training Approach
We propose a new learning to rank algorithm, named Weighted Margin-Rank Batch loss (WMRB), to extend the popular Weighted Approximate-Rank Pairwise loss (WARP). WMRB uses a new rank estimator and an efficient batch training algorithm. The approach allows more accurate item rank approximation and explicit utilization of parallel computation to accelerate training. In three item recommendation ta...
متن کاملA Scalable Kernel-Based Algorithm for Semi-Supervised Metric Learning
In recent years, metric learning in the semisupervised setting has aroused a lot of research interests. One type of semi-supervised metric learning utilizes supervisory information in the form of pairwise similarity or dissimilarity constraints. However, most methods proposed so far are either limited to linear metric learning or unable to scale up well with the data set size. In this paper, we...
متن کاملConvex Perturbations for Scalable Semidefinite Programming
Many important machine learning problems are modeled and solved via semidefinite programs; examples include metric learning, nonlinear embedding, and certain clustering problems. Often, off-the-shelf software is invoked for the associated optimization, which can be inappropriate due to excessive computational and storage requirements. In this paper, we introduce the use of convex perturbations ...
متن کاملRecent process on structured low-rank approximation
Rank deficiency of a data matrix is equivalent to the existence of an exact linear model for the data. For the purpose of linear static modeling, the matrix is unstructured and the corresponding modeling problem is an approximation of the matrix by another matrix of a lower rank. In the context of linear time-invariant dynamic models, the appropriate data matrix is Hankel and the corresponding ...
متن کاملRecent progress in structured low-rank approximation
Rank deficiency of a data matrix is equivalent to the existence of an exact linear model for the data. For the purpose of linear static modeling, the matrix is unstructured and the correspondingmodeling problem is an approximation of the matrix by another matrix of a lower rank. In the context of linear time-invariant dynamic models, the appropriate data matrix is Hankel and the corresponding m...
متن کامل